Hybrid control for combining model-based and model-free reinforcement learning
نویسندگان
چکیده
We develop an approach to improve the learning capabilities of robotic systems by combining learned predictive models with experience-based state-action policy mappings. Predictive provide understanding task and dynamics, while (model-free) mappings encode favorable actions that override planned actions. refer our systematically model-based model-free methods as hybrid learning. Our efficiently learns motor skills improves performance policies. Moreover, enables policies (both model-free) be updated using any off-policy reinforcement method. derive a deterministic method optimally switching between modalities. adapt stochastic variation relaxes some key assumptions in original derivation. variations are tested on variety robot control benchmark tasks simulation well hardware manipulation task. extend for use imitation methods, where experience is provided through demonstrations, we test expanded capability real-world pick-and-place The results show capable improving sample efficiency experimental domains.
منابع مشابه
Combining Model-Based and Model-Free Updates for Trajectory-Centric Reinforcement Learning
Reinforcement learning (RL) algorithms for realworld robotic applications need a data-efficient learning process and the ability to handle complex, unknown dynamical systems. These requirements are handled well by model-based and model-free RL approaches, respectively. In this work, we aim to combine the advantages of these two types of methods in a principled manner. By focusing on time-varyin...
متن کاملCombining Model-Based and Model-Free Updates for Deep Reinforcement Learning
The ability to learn motor skills autonomously is one of the main requirements for deploying robots in unstructured realworld environments. The goal of reinforcement learning (RL) is to learn such skills through trial and error, thus avoiding tedious manual engineering. However, real-world applications of RL have to contend with two often opposing requirements: data-efficient learning and the a...
متن کاملMBMF: Model-Based Priors for Model-Free Reinforcement Learning
Reinforcement Learning is divided in two main paradigms: model-free and model-based. Each of these two paradigms has strengths and limitations, and has been successfully applied to real world domains that are appropriate to its corresponding strengths. In this paper, we present a new approach aimed at bridging the gap between these two paradigms. We aim to take the best of the two paradigms and...
متن کاملModel-Based Value Expansion for Efficient Model-Free Reinforcement Learning
Recent model-free reinforcement learning algorithms have proposed incorporating learned dynamics models as a source of additional data with the intention of reducing sample complexity. Such methods hold the promise of incorporating imagined data coupled with a notion of model uncertainty to accelerate the learning of continuous control tasks. Unfortunately, they rely on heuristics that limit us...
متن کاملReinforcement Learning: Model-free
Simply put, reinforcement learning (RL) is a term used to indicate a large family of dierent algorithms RL that all share two key properties. First, the objective of RL is to learn appropriate behavior through trialand-error experience in a task. Second, in RL, the feedback available to the learning agent is restricted to a reward signal that indicates how well the agent is behaving, but does ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: The International Journal of Robotics Research
سال: 2022
ISSN: ['1741-3176', '0278-3649']
DOI: https://doi.org/10.1177/02783649221083331